Improved Architectures and Training Algorithms for Deep Operator Networks
نویسندگان
چکیده
Operator learning techniques have recently emerged as a powerful tool for maps between infinite-dimensional Banach spaces. Trained under appropriate constraints, they can also be effective in the solution operator of partial differential equations (PDEs) an entirely self-supervised manner. In this work we analyze training dynamics deep networks (DeepONets) through lens Neural Tangent Kernel theory, and reveal bias that favors approximation functions with larger magnitudes. To correct propose to adaptively re-weight importance each example, demonstrate how procedure effectively balance magnitude back-propagated gradients during via gradient descent. We novel network architecture is more resilient vanishing pathologies. Taken together, our developments provide new insights into DeepONets consistently improve their predictive accuracy by factor 10-50x, demonstrated challenging setting PDE operators absence paired input-output observations. All code data accompanying manuscript will made publicly available at https://github.com/PredictiveIntelligenceLab/ImprovedDeepONets .
منابع مشابه
Implementation of Evolutionary Algorithms for Deep Architectures
Deep learning is becoming an increasingly interesting and powerful machine learning method with successful applications in many domains, such as natural language processing, image recognition, and hand-written character recognition. Despite of its eminent success, limitations of traditional learning approach may still prevent deep learning from achieving a wide range of realistic learning tasks...
متن کاملDeepArchitect: Automatically Designing and Training Deep Architectures
In deep learning, performance is strongly affected by the choice of architecture and hyperparameters. While there has been extensive work on automatic hyperparameter optimization for simple spaces, complex spaces such as the space of deep architectures remain largely unexplored. As a result, the choice of architecture is done manually by the human expert through a slow trial and error process g...
متن کاملDeep Neuroevolution: Genetic Algorithms Are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning
Deep artificial neural networks (DNNs) are typically trained via gradient-based learning algorithms, namely backpropagation. Evolution strategies (ES) can rival backprop-based algorithms such as Q-learning and policy gradients on challenging deep reinforcement learning (RL) problems. However, ES can be considered a gradient-based algorithm because it performs stochastic gradient descent via an ...
متن کاملDeepDefense: Training Deep Neural Networks with Improved Robustness
Despite the efficacy on a variety of computer vision tasks, deep neural networks (DNNs) are vulnerable to adversarial attacks, limiting their applications in security-critical systems. Recent works have shown the possibility of generating imperceptibly perturbed image inputs (a.k.a., adversarial examples) to fool well-trained DNN models into making arbitrary predictions. To address this problem...
متن کاملStable Architectures for Deep Neural Networks
Deep neural networks have become valuable tools for supervised machine learning, e.g., in the classification of text or images. While offering superior results over traditional techniques to find and express complicated patterns in data, deep architectures are known to be challenging to design and train such that they generalize well to new data. An important issue that must be overcome is nume...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of Scientific Computing
سال: 2022
ISSN: ['1573-7691', '0885-7474']
DOI: https://doi.org/10.1007/s10915-022-01881-0